19 research outputs found

    Generative Pre-trained Transformer for Vietnamese Community-based COVID-19 Question Answering

    Full text link
    Recent studies have provided empirical evidence of the wide-ranging potential of Generative Pre-trained Transformer (GPT), a pretrained language model, in the field of natural language processing. GPT has been effectively employed as a decoder within state-of-the-art (SOTA) question answering systems, yielding exceptional performance across various tasks. However, the current research landscape concerning GPT's application in Vietnamese remains limited. This paper aims to address this gap by presenting an implementation of GPT-2 for community-based question answering specifically focused on COVID-19 related queries in Vietnamese. We introduce a novel approach by conducting a comparative analysis of different Transformers vs SOTA models in the community-based COVID-19 question answering dataset. The experimental findings demonstrate that the GPT-2 models exhibit highly promising outcomes, outperforming other SOTA models as well as previous community-based COVID-19 question answering models developed for Vietnamese

    BARTPhoBEiT: Pre-trained Sequence-to-Sequence and Image Transformers Models for Vietnamese Visual Question Answering

    Full text link
    Visual Question Answering (VQA) is an intricate and demanding task that integrates natural language processing (NLP) and computer vision (CV), capturing the interest of researchers. The English language, renowned for its wealth of resources, has witnessed notable advancements in both datasets and models designed for VQA. However, there is a lack of models that target specific countries such as Vietnam. To address this limitation, we introduce a transformer-based Vietnamese model named BARTPhoBEiT. This model includes pre-trained Sequence-to-Sequence and bidirectional encoder representation from Image Transformers in Vietnamese and evaluates Vietnamese VQA datasets. Experimental results demonstrate that our proposed model outperforms the strong baseline and improves the state-of-the-art in six metrics: Accuracy, Precision, Recall, F1-score, WUPS 0.0, and WUPS 0.9

    ViCLEVR: A Visual Reasoning Dataset and Hybrid Multimodal Fusion Model for Visual Question Answering in Vietnamese

    Full text link
    In recent years, Visual Question Answering (VQA) has gained significant attention for its diverse applications, including intelligent car assistance, aiding visually impaired individuals, and document image information retrieval using natural language queries. VQA requires effective integration of information from questions and images to generate accurate answers. Neural models for VQA have made remarkable progress on large-scale datasets, with a primary focus on resource-rich languages like English. To address this, we introduce the ViCLEVR dataset, a pioneering collection for evaluating various visual reasoning capabilities in Vietnamese while mitigating biases. The dataset comprises over 26,000 images and 30,000 question-answer pairs (QAs), each question annotated to specify the type of reasoning involved. Leveraging this dataset, we conduct a comprehensive analysis of contemporary visual reasoning systems, offering valuable insights into their strengths and limitations. Furthermore, we present PhoVIT, a comprehensive multimodal fusion that identifies objects in images based on questions. The architecture effectively employs transformers to enable simultaneous reasoning over textual and visual data, merging both modalities at an early model stage. The experimental findings demonstrate that our proposed model achieves state-of-the-art performance across four evaluation metrics. The accompanying code and dataset have been made publicly accessible at \url{https://github.com/kvt0012/ViCLEVR}. This provision seeks to stimulate advancements within the research community, fostering the development of more multimodal fusion algorithms, specifically tailored to address the nuances of low-resource languages, exemplified by Vietnamese.Comment: A pre-print version and submitted to journa

    Facile Template In-Situ Fabrication of ZnCo2O4 Nanoparticles with Highly Photocatalytic Activities under Visible-Light Irradiation

    Get PDF
    High specific surface area ZnCo2O4 nanoparticles were prepared via a sacrificial template accelerated hydrolysis by using nanoparticles of ZnO with highly polar properties as a template. The obtained ZnCo2O4 nanoparticles were characterized by the method of scanning electron microscopy (SEM), X-ray diffraction (XRD), Brunauer-Emmett-Teller (BET) surface area measurements, Transmission electron microscopy (TEM), Raman spectroscopy, and X-ray photoelectron spectroscopy (XPS). The obtained nanoparticles were performed as a photocatalyst for the degradation of methylene blue in aqueous solution under visible irradiation. The photocatalytic degradation rate of methylene blue onto the synthesized ZnCo2O4 was higher than that of commercial ZnO and synthesized ZnO template. Copyright © 2019 BCREC Group. All rights reserved

    TextANIMAR: Text-based 3D Animal Fine-Grained Retrieval

    Full text link
    3D object retrieval is an important yet challenging task, which has drawn more and more attention in recent years. While existing approaches have made strides in addressing this issue, they are often limited to restricted settings such as image and sketch queries, which are often unfriendly interactions for common users. In order to overcome these limitations, this paper presents a novel SHREC challenge track focusing on text-based fine-grained retrieval of 3D animal models. Unlike previous SHREC challenge tracks, the proposed task is considerably more challenging, requiring participants to develop innovative approaches to tackle the problem of text-based retrieval. Despite the increased difficulty, we believe that this task has the potential to drive useful applications in practice and facilitate more intuitive interactions with 3D objects. Five groups participated in our competition, submitting a total of 114 runs. While the results obtained in our competition are satisfactory, we note that the challenges presented by this task are far from being fully solved. As such, we provide insights into potential areas for future research and improvements. We believe that we can help push the boundaries of 3D object retrieval and facilitate more user-friendly interactions via vision-language technologies.Comment: arXiv admin note: text overlap with arXiv:2304.0573

    Ventilator-associated respiratory infection in a resource-restricted setting: impact and etiology.

    Get PDF
    BACKGROUND: Ventilator-associated respiratory infection (VARI) is a significant problem in resource-restricted intensive care units (ICUs), but differences in casemix and etiology means VARI in resource-restricted ICUs may be different from that found in resource-rich units. Data from these settings are vital to plan preventative interventions and assess their cost-effectiveness, but few are available. METHODS: We conducted a prospective observational study in four Vietnamese ICUs to assess the incidence and impact of VARI. Patients ≥ 16 years old and expected to be mechanically ventilated > 48 h were enrolled in the study and followed daily for 28 days following ICU admission. RESULTS: Four hundred fifty eligible patients were enrolled over 24 months, and after exclusions, 374 patients' data were analyzed. A total of 92/374 cases of VARI (21.7/1000 ventilator days) were diagnosed; 37 (9.9%) of these met ventilator-associated pneumonia (VAP) criteria (8.7/1000 ventilator days). Patients with any VARI, VAP, or VARI without VAP experienced increased hospital and ICU stay, ICU cost, and antibiotic use (p < 0.01 for all). This was also true for all VARI (p < 0.01 for all) with/without tetanus. There was no increased risk of in-hospital death in patients with VARI compared to those without (VAP HR 1.58, 95% CI 0.75-3.33, p = 0.23; VARI without VAP HR 0.40, 95% CI 0.14-1.17, p = 0.09). In patients with positive endotracheal aspirate cultures, most VARI was caused by Gram-negative organisms; the most frequent were Acinetobacter baumannii (32/73, 43.8%) Klebsiella pneumoniae (26/73, 35.6%), and Pseudomonas aeruginosa (24/73, 32.9%). 40/68 (58.8%) patients with positive cultures for these had carbapenem-resistant isolates. Patients with carbapenem-resistant VARI had significantly greater ICU costs than patients with carbapenem-susceptible isolates (6053 USD (IQR 3806-7824) vs 3131 USD (IQR 2108-7551), p = 0.04) and after correction for adequacy of initial antibiotics and APACHE II score, showed a trend towards increased risk of in-hospital death (HR 2.82, 95% CI 0.75-6.75, p = 0.15). CONCLUSIONS: VARI in a resource-restricted setting has limited impact on mortality, but shows significant association with increased patient costs, length of stay, and antibiotic use, particularly when caused by carbapenem-resistant bacteria. Evidence-based interventions to reduce VARI in these settings are urgently needed

    UIT-HSE at WNUT-2020 Task 2: Exploiting CT-BERT for Identifying COVID-19 Information on the Twitter Social Network

    No full text
    Recently, COVID-19 has affected a variety of real-life aspects of the world and led to dreadful consequences. More and more tweets about COVID-19 has been shared publicly on Twitter. However, the plurality of those Tweets are uninformative, which is challenging to build automatic systems to detect the informative ones for useful AI applications. In this paper, we present our results at the W-NUT 2020 Shared Task 2: Identification of Informative COVID-19 English Tweets. In particular, we propose our simple but effective approach using the transformer-based models based on COVID-Twitter-BERT (CT-BERT) with different fine-tuning techniques. As a result, we achieve the F1-Score of 90.94\% with the third place on the leaderboard of this task which attracted 56 submitted teams in total
    corecore